6
Introduction
FIGURE 1.1
Modulation process based on an M-Filter.
(e.g., sign function) for the reconstructed weights. While the reconstruction is binarized,
the computation in the latent factorized space is done in the real domain. This has several
advantages. First, the latent factorization enforces a coupling of filters before binarization,
which significantly improves the accuracy of trained models. Second, during training, the
binary weights of each convolutional layer are parametrized using a real-valued matrix or
tensor decomposition, while during inference, reconstructed (binary) weights are used.
Instead of using the same binary method for weights and activations, Huang et al. [93]
believe that the best performance for binarized neural networks can be obtained by applying
different quantization methods to weights and activations. They simultaneously binarize the
weights and quantize the activations to reduce bandwidth.
ReActNet [158] proposes a simple channel-wise reshaping and shifting operation for the
activation distribution, which replaces the sign function with ReAct-Sign, and replaces the
PReLU function with ReAct-PReLU. The parameters in ReAct-Sign and ReAct-PReLU
can be updated.
Compared to XNOR-Net [199], both HORQ-Net [138] and ABC-Net [147] use mul-
tiple binary weights and activations. As a result, HORQ-Net and ABC-Net outperform
XNOR-Net on binary tasks, but they also increase complexity, which goes against the ini-
tial intention of BNNs. New neural networks that perform better and retain the advantage
of speediness are waiting to be explored. MCN [236] and LBCNN [109] proposed new filters
while quantizing parameters and introducing a new loss function to learn these auxiliary
filters.
1.1.4
Structural Design
The basic structure of networks such as BinaryConnect [48] and BinaryNet [99] is essentially
the same as traditional CNNs, which may not fit the binary process. Some attempts have
been made to modify the structure of BNNs for better accuracy.
FIGURE 1.2
MCNs convolution.